9 research outputs found

    SpaceNet MVOI: a Multi-View Overhead Imagery Dataset

    Full text link
    Detection and segmentation of objects in overheard imagery is a challenging task. The variable density, random orientation, small size, and instance-to-instance heterogeneity of objects in overhead imagery calls for approaches distinct from existing models designed for natural scene datasets. Though new overhead imagery datasets are being developed, they almost universally comprise a single view taken from directly overhead ("at nadir"), failing to address a critical variable: look angle. By contrast, views vary in real-world overhead imagery, particularly in dynamic scenarios such as natural disasters where first looks are often over 40 degrees off-nadir. This represents an important challenge to computer vision methods, as changing view angle adds distortions, alters resolution, and changes lighting. At present, the impact of these perturbations for algorithmic detection and segmentation of objects is untested. To address this problem, we present an open source Multi-View Overhead Imagery dataset, termed SpaceNet MVOI, with 27 unique looks from a broad range of viewing angles (-32.5 degrees to 54.0 degrees). Each of these images cover the same 665 square km geographic extent and are annotated with 126,747 building footprint labels, enabling direct assessment of the impact of viewpoint perturbation on model performance. We benchmark multiple leading segmentation and object detection models on: (1) building detection, (2) generalization to unseen viewing angles and resolutions, and (3) sensitivity of building footprint extraction to changes in resolution. We find that state of the art segmentation and object detection models struggle to identify buildings in off-nadir imagery and generalize poorly to unseen views, presenting an important benchmark to explore the broadly relevant challenge of detecting small, heterogeneous target objects in visually dynamic contexts.Comment: Accepted into IEEE International Conference on Computer Vision (ICCV) 201

    Mapping Crop Residue by Combining Landsat and WorldView-3 Satellite Imagery

    No full text
    A unique, multi-tiered approach was applied to map crop residue cover on the Eastern Shore of the Chesapeake Bay, United States. Field measurements of crop residue cover were used to calibrate residue mapping using shortwave infrared (SWIR) indices derived from WorldView-3 imagery for a 12-km × 12-km footprint. The resulting map was then used to calibrate and subsequently classify crop residue mapping using Landsat imagery at a larger spatial resolution and extent. This manuscript describes how the method was applied and presents results in the form of crop residue cover maps, validation statistics, and quantification of conservation tillage implementation in the agricultural landscape. Overall accuracy for maps derived from Landsat 7 and Landsat 8 were comparable at roughly 92% (+/− 10%). Tillage class-specific accuracy was also strong and ranged from 75% to 99%. The approach, which employed a 12-band image stack of six tillage spectral indices and six individual Landsat bands, was shown to be adaptable to variable soil moisture conditions—under dry conditions (Landsat 7, 14 May 2015) the majority of predictive power was attributed to SWIR indices, and under wet conditions (Landsat 8, 22 May 2015) single band reflectance values were more effective at explaining variability in residue cover. Summary statistics of resulting tillage class occurrence matched closely with conservation tillage implementation totals reported by Maryland and Delaware to the Chesapeake Bay Program. This hybrid method combining WorldView-3 and Landsat imagery sources shows promise for monitoring progress in the adoption of conservation tillage practices and for describing crop residue outcomes associated with a variety of agricultural management practices

    Mapping Crop Residue and Tillage Intensity Using WorldView-3 Satellite Shortwave Infrared Residue Indices

    No full text
    Crop residues serve many important functions in agricultural conservation including preserving soil moisture, building soil organic carbon, and preventing erosion. Percent crop residue cover on a field surface reflects the outcome of tillage intensity and crop management practices. Previous studies using proximal hyperspectral remote sensing have demonstrated accurate measurement of percent residue cover using residue indices that characterize cellulose and lignin absorption features found between 2100 nm and 2300 nm in the shortwave infrared (SWIR) region of the electromagnetic spectrum. The 2014 launch of the WorldView-3 (WV3) satellite has now provided a space-borne platform for the collection of narrow band SWIR reflectance imagery capable of measuring these cellulose and lignin absorption features. In this study, WorldView-3 SWIR imagery (14 May 2015) was acquired over farmland on the Eastern Shore of Chesapeake Bay (Maryland, USA), was converted to surface reflectance, and eight different SWIR reflectance indices were calculated. On-farm photographic sampling was used to measure percent residue cover at a total of 174 locations in 10 agricultural fields, ranging from plow-till to continuous no-till management, and these in situ measurements were used to develop percent residue cover prediction models from the SWIR indices using both polynomial and linear least squares regressions. Analysis was limited to agricultural fields with minimal green vegetation (Normalized Difference Vegetation Index < 0.3) due to expected interference of vegetation with the SWIR indices. In the resulting residue prediction models, spectrally narrow residue indices including the Shortwave Infrared Normalized Difference Residue Index (SINDRI) and the Lignin Cellulose Absorption Index (LCA) were determined to be more accurate than spectrally broad Landsat-compatible indices such as the Normalized Difference Tillage Index (NDTI), as determined by respective R2 values of 0.94, 0.92, and 0.84 and respective residual mean squared errors (RMSE) of 7.15, 8.40, and 12.00. Additionally, SINDRI and LCA were more resistant to interference from low levels of green vegetation. The model with the highest correlation (2nd order polynomial SINDRI, R2 = 0.94) was used to convert the SWIR imagery into a map of crop residue cover for non-vegetated agricultural fields throughout the imagery extent, describing the distribution of tillage intensity within the farm landscape. WorldView-3 satellite imagery provides spectrally narrow SWIR reflectance measurements that show utility for a robust mapping of crop residue cover

    SpaceNet 6: Multi-Sensor All Weather Mapping Dataset

    No full text
    Within the remote sensing domain, a diverse set of acquisition modalities exist, each with their own unique strengths and weaknesses. Yet, most of the current literature and open datasets only deal with electro-optical (optical) data for different detection and segmentation tasks at high spatial resolutions. optical data is often the preferred choice for geospatial applications, but requires clear skies and little cloud cover to work well. Conversely, Synthetic Aperture Radar (SAR) sensors have the unique capability to penetrate clouds and collect during all weather, day and night conditions. Consequently, SAR data are particularly valuable in the quest to aid disaster response, when weather and cloud cover can obstruct traditional optical sensors. Despite all of these advantages, there is little open data available to researchers to explore the effectiveness of SAR for such applications, particularly at very-high spatial resolutions, i.e. < 1m Ground Sample Distance (GSD). To address this problem, we present an open MultiSensor All Weather Mapping (MSAW) dataset and challenge, which features two collection modalities (both SAR and optical). The dataset and challenge focus on mapping and building footprint extraction using a combination of these data sources. MSAW covers 120km2 over multiple overlapping collects and is annotated with over 48, 000 unique building footprints labels, enabling the creation and evaluation of mapping algorithms for multi-modal data. We present a baseline and benchmark for building footprint extraction with SAR data and find that state-of-the-art segmentation models pre-trained on optical data, and then trained on SAR (F1 score of 0.21) outperform those trained on SAR data alone (F1 score of 0.135)
    corecore